skip to main content


Search for: All records

Creators/Authors contains: "Mehta, Piyush M."

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Transform and entropy models are the two core components in deep image compression neural networks. Most existing learning-based image compression methods utilize convolutional-based transform, which lacks the ability to model long-range dependencies, primarily due to the limited receptive field of the convolution operation. To address this limitation, we propose a Transformer-based nonlinear transform. This transform has the remarkable ability to efficiently capture both local and global information from the input image, leading to a more decorrelated latent representation. In addition, we introduce a novel entropy model that incorporates two different hyperpriors to model cross-channel and spatial dependencies of the latent representation. To further improve the entropy model, we add a global context that leverages distant relationships to predict the current latent more accurately. This global context employs a causal attention mechanism to extract long-range information in a content-dependent manner. Our experiments show that our proposed framework performs better than the state-of-the-art methods in terms of rate-distortion performance. 
    more » « less
    Free, publicly-accessible full text available December 15, 2024
  2. Abstract

    Space weather indices are used commonly to drive forecasts of thermosphere density, which affects objects in low‐Earth orbit (LEO) through atmospheric drag. One commonly used space weather proxy,F10.7cm, correlates well with solar extreme ultra‐violet (EUV) energy deposition into the thermosphere. Currently, the USAF contracts Space Environment Technologies (SET), which uses a linear algorithm to forecastF10.7cm. In this work, we introduce methods using neural network ensembles with multi‐layer perceptrons (MLPs) and long‐short term memory (LSTMs) to improve on the SET predictions. We make predictions only from historicalF10.7cmvalues. We investigate data manipulation methods (backwards averaging and lookback) as well as multi step and dynamic forecasting. This work shows an improvement over the popular persistence and the operational SET model when using ensemble methods. The best models found in this work are ensemble approaches using multi step or a combination of multi step and dynamic predictions. Nearly all approaches offer an improvement, with the best models improving between 48% and 59% on relative MSE with respect to persistence. Other relative error metrics were shown to improve greatly when ensembles methods were used. We were also able to leverage the ensemble approach to provide a distribution of predicted values; allowing an investigation into forecast uncertainty. Our work found models that produced less biased predictions at elevated and high solar activity levels. Uncertainty was also investigated through the use of a calibration error score metric (CES), our best ensemble reached similar CES as other work.

     
    more » « less
  3. Abstract

    The geospace environment is volatile and highly driven. Space weather has effects on Earth's magnetosphere that cause a dynamic and enigmatic response in the thermosphere, particularly on the evolution of neutral mass density. Many models exist that use space weather drivers to produce a density response, but these models are typically computationally expensive or inaccurate for certain space weather conditions. In response, this work aims to employ a probabilistic machine learning (ML) method to create an efficient surrogate for the Thermosphere Ionosphere Electrodynamics General Circulation Model (TIE‐GCM), a physics‐based thermosphere model. Our method leverages principal component analysis to reduce the dimensionality of TIE‐GCM and recurrent neural networks to model the dynamic behavior of the thermosphere much quicker than the numerical model. The newly developed reduced order probabilistic emulator (ROPE) uses Long‐Short Term Memory neural networks to perform time‐series forecasting in the reduced state and provide distributions for future density. We show that across the available data, TIE‐GCM ROPE has similar error to previous linear approaches while improving storm‐time modeling. We also conduct a satellite propagation study for the significant November 2003 storm which shows that TIE‐GCM ROPE can capture the position resulting from TIE‐GCM density with <5 km bias. Simultaneously, linear approaches provide point estimates that can result in biases of 7–18 km.

     
    more » « less
  4. Abstract

    Machine learning (ML) has been applied to space weather problems with increasing frequency in recent years, driven by an influx of in-situ measurements and a desire to improve modeling and forecasting capabilities throughout the field. Space weather originates from solar perturbations and is comprised of the resulting complex variations they cause within the numerous systems between the Sun and Earth. These systems are often tightly coupled and not well understood. This creates a need for skillful models with knowledge about the confidence of their predictions. One example of such a dynamical system highly impacted by space weather is the thermosphere, the neutral region of Earth’s upper atmosphere. Our inability to forecast it has severe repercussions in the context of satellite drag and computation of probability of collision between two space objects in low Earth orbit (LEO) for decision making in space operations. Even with (assumed) perfect forecast of model drivers, our incomplete knowledge of the system results in often inaccurate thermospheric neutral mass density predictions. Continuing efforts are being made to improve model accuracy, but density models rarely provide estimates of confidence in predictions. In this work, we propose two techniques to develop nonlinear ML regression models to predict thermospheric density while providing robust and reliable uncertainty estimates: Monte Carlo (MC) dropout and direct prediction of the probability distribution, both using the negative logarithm of predictive density (NLPD) loss function. We show the performance capabilities for models trained on both local and global datasets. We show that the NLPD loss provides similar results for both techniques but the direct probability distribution prediction method has a much lower computational cost. For the global model regressed on the Space Environment Technologies High Accuracy Satellite Drag Model (HASDM) density database, we achieve errors of approximately 11% on independent test data with well-calibrated uncertainty estimates. Using an in-situ CHAllenging Minisatellite Payload (CHAMP) density dataset, models developed using both techniques provide test error on the order of 13%. The CHAMP models—on validation and test data—are within 2% of perfect calibration for the twenty prediction intervals tested. We show that this model can also be used to obtain global density predictions with uncertainties at a given epoch.

     
    more » « less
  5. Abstract

    We present results from a study of the time lags between changes in the energy flow into the polar regions and the response of the thermosphere to the heating. Measurements of the neutral density from the Challenging Mini‐satellite Payload (CHAMP) and Gravity Recovery and Climate Experiment (GRACE) missions are used, along with calculations of the total Poynting flux entering the poles. During two major geomagnetic storms in 2003, these data show increased densities are first seen on the dayside edge of the auroral ovals after a surge in the energy input. At lower latitudes, the densities reach their peak values on the dayside earlier than on the night side. A puzzling response seen in the CHAMP measurements during the November 2003 storm was that the density at a fixed location near the “Harang discontinuity” remained at unusually low levels during three sequential orbit passes, while elsewhere the density increased. The entire database of measurements from the CHAMP and GRACE missions were used to derive maps of the density time lags across the globe. The maps show a large gradient between short and long time delays between 60° and 30° geographic latitude. They confirm the findings from the two storm periods, that near the equator, the density on the dayside responds earlier than on the nightside. The time lags are longest near 18–20 hr local time. The time lag maps could be applied to improve the accuracy of empirical thermosphere models, and developers of numerical models may find these results useful for comparisons with their calculations.

     
    more » « less
  6. Abstract

    The Mass Spectrometer and Incoherent Scatter radar (MSIS) model family has been developed and improved since the early 1970's. The most recent version of MSIS is the Naval Research Laboratory (NRL) MSIS 2.0 empirical atmospheric model. NRLMSIS 2.0 provides species density, mass density, and temperature estimates as function of location and space weather conditions. MSIS models have long been a popular choice of thermosphere model in the research and operations community alike, but—like many models—does not provide uncertainty estimates. In this work, we develop an exospheric temperature model based in machine learning that can be used with NRLMSIS 2.0 to calibrate it relative to high‐fidelity satellite density estimates directly through the exospheric temperature parameter. Instead of providing point estimates, our model (called MSIS‐UQ) outputs a distribution which is assessed using a metric called the calibration error score. We show that MSIS‐UQ debiases NRLMSIS 2.0 resulting in reduced differences between model and satellite density of 25% and is 11% closer to satellite density than the Space Force's High Accuracy Satellite Drag Model. We also show the model's uncertainty estimation capabilities by generating altitude profiles for species density, mass density, and temperature. This explicitly demonstrates how exospheric temperature probabilities affect density and temperature profiles within NRLMSIS 2.0. Another study displays improved post‐storm overcooling capabilities relative to NRLMSIS 2.0 alone, enhancing the phenomena that it can capture.

     
    more » « less
  7. Abstract

    Machine learning (ML) models are universal function approximators and—if used correctly—can summarize the information content of observational data sets in a functional form for scientific and engineering applications. A benefit to ML over parametric models is that there are no a priori assumptions about particular basis functions which can potentially limit the phenomena that can be modeled. In this work, we develop ML models on three data sets: the Space Environment Technologies High Accuracy Satellite Drag Model (HASDM) density database, a spatiotemporally matched data set of outputs from the Jacchia‐Bowman 2008 Empirical Thermospheric Density Model (JB2008), and an accelerometer‐derived density data set from CHAllenging Minisatellite Payload (CHAMP). These ML models are compared to the Naval Research Laboratory Mass Spectrometer and Incoherent Scatter radar (NRLMSIS 2.0) model to study the presence of post‐storm cooling in the middle‐thermosphere. We find that both NRLMSIS 2.0 and JB2008‐ML do not account for post‐storm cooling and consequently perform poorly in periods following strong geomagnetic storms (e.g., the 2003 Halloween storms). Conversely, HASDM‐ML and CHAMP‐ML do show evidence of post‐storm cooling indicating that this phenomenon is present in the original data sets. Results show that density reductions up to 40% can occur 1–3 days post‐storm depending on the location and strength of the storm.

     
    more » « less
  8. Abstract

    The EXospheric TEMperatures on a PoLyhedrAl gRid (EXTEMPLAR) method predicts the neutral densities in the thermosphere. The performance of this model has been evaluated through a comparison with the Air Force High Accuracy Satellite Drag Model (HASDM). The Space Environment Technologies (SET) HASDM database that was used for this test spans the 20 years 2000 through 2019, containing densities at 3 hr time intervals at 25 km altitude steps, and a spatial resolution of 10° latitude by 15° longitude. The upgraded EXTEMPLAR that was tested uses the newer Naval Research Laboratory MSIS 2.0 model to convert global exospheric temperature values to neutral density as a function of altitude. The revision also incorporated time delays that varied as a function of location, between the total Poynting flux in the polar regions and the exospheric temperature response. The density values from both models were integrated on spherical shells at altitudes ranging from 200 to 800 km. These sums were compared as a function of time. The results show an excellent agreement at temporal scales ranging from hours to years. The EXTEMPLAR model performs best at altitudes of 400 km and above, where geomagnetic storms produce the largest relative changes in neutral density. In addition to providing an effective method to compare models that have very different spatial resolutions, the use of density totals at various altitudes presents a useful illustration of how the thermosphere behaves at different altitudes, on time scales ranging from hours to complete solar cycles.

     
    more » « less